Input Feature Selection Method Based on Feature Set Equivalence and Mutual Information Gain Maximization

نویسندگان
چکیده

برای دانلود باید عضویت طلایی داشته باشید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Input Feature Selection by Mutual Information Based on Parzen Window

Mutual information is a good indicator of relevance between variables, and have been used as a measure in several feature selection algorithms. However, calculating the mutual information is difficult, and the performance of a feature selection algorithm depends on the accuracy of the mutual information. In this paper, we propose a new method of calculating mutual information between input and ...

متن کامل

Feature Selection Based on Joint Mutual Information

A feature/input selection method is proposed based on joint mutual information. The new method is better than the existing methods based on mutual information in eliminating redundancy in the inputs. It is applied in a real world application to nd 2-D viewing coordinates for data visualization and to select inputs for a neural network classiier. The result shows that the new method can nd many ...

متن کامل

Feature Selection based on Information Gain

The attribute reduction is one of the key processes for knowledge acquisition. Some data set is multidimensional and larger in size. If that data set is used for classification it may end with wrong results and it may also occupy more resources especially in terms of time. Most of the features present are redundant and inconsistent and affect the classification. In order to improve the efficien...

متن کامل

MIFS-ND: A mutual information-based feature selection method

Feature selection is used to choose a subset of relevant features for effective classification of data. In high dimensional data classification, the performance of a classifier often depends on the feature subset used for classification. In this paper, we introduce a greedy feature selection method using mutual information. This method combines both feature–feature mutual information and featur...

متن کامل

Conditional Dynamic Mutual Information-Based Feature Selection

With emergence of new techniques, data in many fields are getting larger and larger, especially in dimensionality aspect. The high dimensionality of data may pose great challenges to traditional learning algorithms. In fact, many of features in large volume of data are redundant and noisy. Their presence not only degrades the performance of learning algorithms, but also confuses end-users in th...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: IEEE Access

سال: 2019

ISSN: 2169-3536

DOI: 10.1109/access.2019.2948095